-
Couldn't load subscription status.
- Fork 1.9k
[DO NOT MERGE]Try to run v1 test using v2 connector #5338
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
huan233usc
wants to merge
55
commits into
delta-io:master
Choose a base branch
from
huan233usc:run-v1-test
base: master
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
…ectory to avoid evaluation order issues
- Changed from baseDirectory.getParentFile/spark to (delta-spark-v1/baseDirectory).value - This is more explicit and clearly shows we're using delta-spark-v1's directory - Makes the relationship between modules more obvious
Issue: Test resource directories were using baseDirectory.getParentFile/spark which could evaluate to the wrong path depending on evaluation order. Solution: Changed all test path configurations to consistently use (delta-spark-v1/baseDirectory).value: - Test/unmanagedSourceDirectories - Test/unmanagedResourceDirectories - Test/resourceDirectory - Test/baseDirectory - Test/javaOptions (-Duser.dir) This ensures all test paths correctly point to the spark/ directory regardless of evaluation order, fixing GitHub Actions failures.
The spark module was adding all test javaOptions again (which are already in commonSettings), causing duplicates. Now it only adds -Duser.dir which is spark-specific.
…ectory Root cause: TestParallelization.defaultForkOptions was using baseDirectory.value for workingDirectory, but spark module's Test/baseDirectory points to spark/ while baseDirectory points to spark-combined/. When GitHub Actions runs 'spark/test' with TEST_PARALLELISM_COUNT=4 SHARD_ID=x, the forked test JVMs got spark-combined/ as working directory, causing tests that use relative paths (like 'src/test/resources/delta/table-with-dv-large') to fail. Solution: Changed defaultForkOptions to use (Test/baseDirectory).value instead of baseDirectory.value, so it correctly uses spark/ as the working directory. This only affects the spark module which is the only user of TestParallelization.
Issue: serverClassPath contains multiple 'classes' directories with the same name (e.g., spark/target/scala-2.12/classes, storage/target/scala-2.12/classes, etc.). When creating symlinks, the code tried to create multiple symlinks all named 'classes', causing FileAlreadyExistsException. Solution: Track created symlink names in a Set and skip duplicates. Only the first occurrence of each filename will have a symlink created. Also added Files.exists() check and similar fix for log4j properties symlink.
The issue is simply that serverClassPath contains multiple directories with the same name (e.g., 7 different 'classes' directories). Using a Set to track created symlink names is sufficient - no need for try-catch or concurrent access handling since each shard runs in its own workspace.
Changed kernelDefaults to depend on local delta-spark-v1 instead of published delta-spark 3.3.2. This makes the dependency consistent with goldenTables (which already uses delta-spark-v1) and allows testing against the current codebase. Changes: - Added .dependsOn(`delta-spark-v1` % "test") to kernelDefaults - Removed external 'io.delta' %% 'delta-spark' % '3.3.2' % 'test' dependency
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Which Delta project/connector is this regarding?
Description
How was this patch tested?
Does this PR introduce any user-facing changes?